本文提出了一个统一的框架到(i)找到球,(ii)预测姿势,(iii)在团队体育场景中分段播放器的实例掩码。这些问题对自动体育分析,生产和广播有高兴趣。常见做法是通过利用通用最先进的模型,例如Panoptic-Deeblab来单独解决每个问题,用于玩家分割。除了从单任务模型的乘法乘以增加的复杂性之外,由于团队体育场景的复杂性和特异性,使用现成的架子模型也会阻碍性能,如强大的遮挡和运动模糊。为了规避这些限制,我们的论文提出培训一种单一的模型,它通过组合零件强度场和空间嵌入原理来预测球和玩家掩模和姿势。部件强度场提供球和播放器位置,以及播放器接头位置。然后利用空间嵌入来将播放器实例像素联系到其各自的播放器中心,而且还将播放器接头分组成骷髅。我们展示了拟议模型在DeepSport篮球数据集上的有效性,为单独解决每个单独任务的SOA模型实现了可比性的性能。
translated by 谷歌翻译
在本文中,我们提出了一个样本复杂性,以从嘈杂的样本中学习单纯形。给出了$ n $的数据集,其中包括i.i.d.样品从$ \ mathbb {r}^k $中的未知任意单纯形上的均匀分布中得出,其中假定样品被任意幅度的加性高斯噪声损坏。我们提出了一种策略,该策略可以输出一个单纯概率,总变化距离为$ \ epsilon + o \ left(\ mathrm {snr}^{ - 1} \ right)$从true Simplex中,对于任何$ \ Epsilon> 0 $。我们证明,要接近True Simplex,就足以拥有$ n \ ge \ tilde {o} \ left(k^2/\ epsilon^2 \ right)$ samples。在这里,SNR代表信噪比,可以看作是单纯形直径与噪声的标准偏差的比率。我们的证明是基于样品压缩技术的最新进步,这些进步已经显示出在高维高斯混合模型中的密度估计的紧密范围方面的承诺。
translated by 谷歌翻译
同工型是从同一基因位点产生的MRNA,称为替代剪接。研究表明,超过95%的人类多外XEX基因经历了替代剪接。尽管mRNA序列的变化很少,但它们可能会对细胞功能和调节产生系统的影响。广泛报道了基因的同工型具有不同甚至对比的功能。大多数研究表明,替代剪接在人类健康和疾病中起着重要作用。尽管具有广泛的基因功能研究,但关于同工型功能的信息很少。最近,已经提出了一些基于多个实例学习的计算方法,用于使用基因函数和基因表达谱预测同工型函数。但是,由于缺乏标记的培训数据,他们的性能并不理想。另外,概率模型(例如条件随机场(CRF))已被用于建模同工型之间的关系。该项目使用所有数据和有价值的信息,例如同工型序列,表达曲线和基因本体论图,并提出了基于深神经网络的综合模型。 Uniprot基因本体论(GO)数据库用作基因函数的标准参考。 NCBI REFSEQ数据库用于提取基因和同工型序列,NCBI SRA数据库用于表达式配置文件数据。曲线下(ROC AUC)下的接收器操作特征区域和曲线下的Precision-Recall等指标用于测量预测准确性。
translated by 谷歌翻译
本文旨在提出和理论上分析一种新的分布式方案,用于稀疏线性回归和特征选择。主要目标是根据来自未知稀疏线性模型的嘈杂观测来了解高维数据集的几个因果特征。但是,在$ \ mathbb {r} ^ p $中包含$ n $ data样本的假定培训集已经在大型网络上分发,以通过极低的带宽链路连接的$ n $客户端。此外,我们考虑渐近配置$ 1 \ ll n \ ll n \ ll p $。为了从整个数据集推断出原因尺寸,我们提出了一种简单但有效的网络中的信息共享方法。在这方面,我们理论上表明,可以可靠地恢复真正的因果特征,其中o的$ o o \ lex(n \ log p \ light)$跨越网络。与将所有样本传输到单个节点(集中式场景)的微小情况相比,这产生了显着降低的通信成本,该沟通成本是需要$ o \ lef(np \右)$传输。诸如ADMM的更复杂的方案仍然具有$ o ox的通信复杂性(NP \右)$。令人惊讶的是,我们的样本复杂性被证明是与每个节点中固定性能测量的最佳集中方法的相同(最多常数因素),而NA \“{i} ve分散技术的最佳集中方法以$线性地增长N $。本文的理论担保是基于Javanmard等人的最近脱叠套索的分析框架。(2019),并由几个在合成和现实世界数据集上进行的几台计算机实验支持。
translated by 谷歌翻译
Real-world robotic grasping can be done robustly if a complete 3D Point Cloud Data (PCD) of an object is available. However, in practice, PCDs are often incomplete when objects are viewed from few and sparse viewpoints before the grasping action, leading to the generation of wrong or inaccurate grasp poses. We propose a novel grasping strategy, named 3DSGrasp, that predicts the missing geometry from the partial PCD to produce reliable grasp poses. Our proposed PCD completion network is a Transformer-based encoder-decoder network with an Offset-Attention layer. Our network is inherently invariant to the object pose and point's permutation, which generates PCDs that are geometrically consistent and completed properly. Experiments on a wide range of partial PCD show that 3DSGrasp outperforms the best state-of-the-art method on PCD completion tasks and largely improves the grasping success rate in real-world scenarios. The code and dataset will be made available upon acceptance.
translated by 谷歌翻译
This paper deals with the problem of statistical and system heterogeneity in a cross-silo Federated Learning (FL) framework where there exist a limited number of Consumer Internet of Things (CIoT) devices in a smart building. We propose a novel Graph Signal Processing (GSP)-inspired aggregation rule based on graph filtering dubbed ``G-Fedfilt''. The proposed aggregator enables a structured flow of information based on the graph's topology. This behavior allows capturing the interconnection of CIoT devices and training domain-specific models. The embedded graph filter is equipped with a tunable parameter which enables a continuous trade-off between domain-agnostic and domain-specific FL. In the case of domain-agnostic, it forces G-Fedfilt to act similar to the conventional Federated Averaging (FedAvg) aggregation rule. The proposed G-Fedfilt also enables an intrinsic smooth clustering based on the graph connectivity without explicitly specified which further boosts the personalization of the models in the framework. In addition, the proposed scheme enjoys a communication-efficient time-scheduling to alleviate the system heterogeneity. This is accomplished by adaptively adjusting the amount of training data samples and sparsity of the models' gradients to reduce communication desynchronization and latency. Simulation results show that the proposed G-Fedfilt achieves up to $3.99\% $ better classification accuracy than the conventional FedAvg when concerning model personalization on the statistically heterogeneous local datasets, while it is capable of yielding up to $2.41\%$ higher accuracy than FedAvg in the case of testing the generalization of the models.
translated by 谷歌翻译
Solute transport in porous media is relevant to a wide range of applications in hydrogeology, geothermal energy, underground CO2 storage, and a variety of chemical engineering systems. Due to the complexity of solute transport in heterogeneous porous media, traditional solvers require high resolution meshing and are therefore expensive computationally. This study explores the application of a mesh-free method based on deep learning to accelerate the simulation of solute transport. We employ Physics-informed Neural Networks (PiNN) to solve solute transport problems in homogeneous and heterogeneous porous media governed by the advection-dispersion equation. Unlike traditional neural networks that learn from large training datasets, PiNNs only leverage the strong form mathematical models to simultaneously solve for multiple dependent or independent field variables (e.g., pressure and solute concentration fields). In this study, we construct PiNN using a periodic activation function to better represent the complex physical signals (i.e., pressure) and their derivatives (i.e., velocity). Several case studies are designed with the intention of investigating the proposed PiNN's capability to handle different degrees of complexity. A manual hyperparameter tuning method is used to find the best PiNN architecture for each test case. Point-wise error and mean square error (MSE) measures are employed to assess the performance of PiNNs' predictions against the ground truth solutions obtained analytically or numerically using the finite element method. Our findings show that the predictions of PiNN are in good agreement with the ground truth solutions while reducing computational complexity and cost by, at least, three orders of magnitude.
translated by 谷歌翻译
The number of international benchmarking competitions is steadily increasing in various fields of machine learning (ML) research and practice. So far, however, little is known about the common practice as well as bottlenecks faced by the community in tackling the research questions posed. To shed light on the status quo of algorithm development in the specific field of biomedical imaging analysis, we designed an international survey that was issued to all participants of challenges conducted in conjunction with the IEEE ISBI 2021 and MICCAI 2021 conferences (80 competitions in total). The survey covered participants' expertise and working environments, their chosen strategies, as well as algorithm characteristics. A median of 72% challenge participants took part in the survey. According to our results, knowledge exchange was the primary incentive (70%) for participation, while the reception of prize money played only a minor role (16%). While a median of 80 working hours was spent on method development, a large portion of participants stated that they did not have enough time for method development (32%). 25% perceived the infrastructure to be a bottleneck. Overall, 94% of all solutions were deep learning-based. Of these, 84% were based on standard architectures. 43% of the respondents reported that the data samples (e.g., images) were too large to be processed at once. This was most commonly addressed by patch-based training (69%), downsampling (37%), and solving 3D analysis tasks as a series of 2D tasks. K-fold cross-validation on the training set was performed by only 37% of the participants and only 50% of the participants performed ensembling based on multiple identical models (61%) or heterogeneous models (39%). 48% of the respondents applied postprocessing steps.
translated by 谷歌翻译
The JPEG standard is widely used in different image processing applications. One of the main components of the JPEG standard is the quantisation table (QT) since it plays a vital role in the image properties such as image quality and file size. In recent years, several efforts based on population-based metaheuristic (PBMH) algorithms have been performed to find the proper QT(s) for a specific image, although they do not take into consideration the user's opinion. Take an android developer as an example, who prefers a small-size image, while the optimisation process results in a high-quality image, leading to a huge file size. Another pitfall of the current works is a lack of comprehensive coverage, meaning that the QT(s) can not provide all possible combinations of file size and quality. Therefore, this paper aims to propose three distinct contributions. First, to include the user's opinion in the compression process, the file size of the output image can be controlled by a user in advance. Second, to tackle the lack of comprehensive coverage, we suggest a novel representation. Our proposed representation can not only provide more comprehensive coverage but also find the proper value for the quality factor for a specific image without any background knowledge. Both changes in representation and objective function are independent of the search strategies and can be used with any type of population-based metaheuristic (PBMH) algorithm. Therefore, as the third contribution, we also provide a comprehensive benchmark on 22 state-of-the-art and recently-introduced PBMH algorithms on our new formulation of JPEG image compression. Our extensive experiments on different benchmark images and in terms of different criteria show that our novel formulation for JPEG image compression can work effectively.
translated by 谷歌翻译
This paper presents a Temporal Graph Neural Network (TGNN) framework for detection and localization of false data injection and ramp attacks on the system state in smart grids. Capturing the topological information of the system through the GNN framework along with the state measurements can improve the performance of the detection mechanism. The problem is formulated as a classification problem through a GNN with message passing mechanism to identify abnormal measurements. The residual block used in the aggregation process of message passing and the gated recurrent unit can lead to improved computational time and performance. The performance of the proposed model has been evaluated through extensive simulations of power system states and attack scenarios showing promising performance. The sensitivity of the model to intensity and location of the attacks and model's detection delay versus detection accuracy have also been evaluated.
translated by 谷歌翻译